Robust regression via error tolerance
نویسندگان
چکیده
Abstract Real-world datasets are often characterised by outliers; data items that do not follow the same structure as rest of data. These outliers might negatively influence modelling In analysis it is, therefore, important to consider methods robust outliers. this paper we develop a regression method finds largest subset can be approximated using sparse linear model given precision. We show yield best possible robustness However, problem is NP-hard and solve present an efficient approximation algorithm, termed SLISE. Our extends existing state-of-the-art methods, especially in terms speed on high-dimensional datasets. demonstrate our applying both synthetic real-world problems.
منابع مشابه
Robust Regression via Hard Thresholding
We study the problem of Robust Least Squares Regression (RLSR) where several response variables can be adversarially corrupted. More specifically, for a data matrix X ∈ Rp×n and an underlying model w∗, the response vector is generated as y = XTw∗+b where b ∈ R is the corruption vector supported over at most C ·n coordinates. Existing exact recovery results for RLSR focus solely on L1-penalty ba...
متن کاملRobust Regression via Hard Thresholding
We study the problem of Robust Least Squares Regression (RLSR) where several response variables can be adversarially corrupted. More specifically, for a data matrix X ∈ Rp×n and an underlying model w∗, the response vector is generated as y = XTw∗ +b where b ∈ R is the corruption vector supported over at most C · n coordinates. Existing exact recovery results for RLSR focus solely on L1penalty b...
متن کاملRelative error prediction via kernel regression smoothers
In this article, we introduce and study local constant and our preferred local linear nonparametric regression estimators when it is appropriate to assess performance in terms of mean squared relative error of prediction. We give asymptotic results for both boundary and non-boundary cases. These are special cases of more general asymptotic results that we provide concerning the estimation of th...
متن کاملRobust Subspace Clustering via Thresholding Ridge Regression
In this material, we provide the theoretical analyses to show that the trivial coefficients always correspond to the codes over errors. Lemmas 1–3 show that our errors-removing strategy will perform well when the lp-norm is enforced over the representation, where p = {1, 2,∞}. Let x 6= 0 be a data point in the union of subspaces SD that is spanned by D = [Dx D−x], where Dx and D−x consist of th...
متن کاملRobust Regression via Heuristic Hard Thresholding
The presence of data noise and corruptions recently invokes increasing attention on Robust Least Squares Regression (RLSR), which addresses the fundamental problem that learns reliable regression coefficients when response variables can be arbitrarily corrupted. Until now, several important challenges still cannot be handled concurrently: 1) exact recovery guarantee of regression coefficients 2...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Data Mining and Knowledge Discovery
سال: 2022
ISSN: ['1573-756X', '1384-5810']
DOI: https://doi.org/10.1007/s10618-022-00819-2